Fog Computing Enabled Locality Based Product Demand Prediction and Decision Making Using Reinforcement Learning

نویسندگان

چکیده

Wastage of perishable and non-perishable products due to manual monitoring in shopping malls creates huge revenue loss supermarket industry. Besides, internal external factors such as calendar events weather condition contribute excess wastage different regions supermarket. It is a challenging job know about the manually supermarkets region-wise. Therefore, management needs take appropriate decision action prevent products. The fog computing data centers located each region can collect, process analyze for demand prediction making. In this paper, product-demand model designed using integrated Principal Component Analysis (PCA) K-means Unsupervised Learning (UL) algorithms making developed State-Action-Reward-State-Action (SARSA) Reinforcement (RL) algorithm. Our proposed method cluster into low, medium, high-demand product by learning from features. Taking derived model, distributing low-demand be made SARSA. Experimental results show that our datasets well with Silhouette score ≥60%. adopted SARSA-based outperforms over Q-Learning, Monte-Carlo, Deep Q-Network (DQN), Actor-Critic terms maximum cumulative reward, average reward execution time.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Stimulus detection and decision making via spike-based reinforcement learning

In theoretical and experimental investigations of decision-making, the main task has typically been one of classification, wherein the relevant stimuli cueing decisions are known to the decision maker: the latter knows which stimuli are relevant, and knows when it is being presented with one. However, in many real-life situations it is not clear which segments in a continuous sensory stream are...

متن کامل

MONEYBaRL: Exploiting pitcher decision-making using Reinforcement Learning

This manuscript uses machine learning techniques to exploit baseball pitchers’ decision making, so-called “Baseball IQ,” by modeling the at-bat information, pitch selection and counts, as a Markov Decision Process (MDP). Each state of the MDP models the pitcher’s current pitch selection in a Markovian fashion, conditional on the information immediately prior to making the current pitch. This in...

متن کامل

Hierarchical Reinforcement Learning and Decision Making for Intelligent Machines

A methodology for performance improvement of Intelligent Machines based on Hierarchical Reinforcement Learning is introduced. Machine Decision Making and Learning are based on a cost function which includes reliability and computational cost of algorithms at the three levels of the hierarchy proposed by Saridis. Despite this particular formalization, the methodology intends to be suuciently gen...

متن کامل

Expectancies in Decision Making, Reinforcement Learning, and Ventral Striatum

Decisions can arise in different ways, such as from a gut feeling, doing what worked last time, or planful deliberation. Different decision-making systems are dissociable behaviorally, map onto distinct brain systems, and have different computational demands. For instance, "model-free" decision strategies use prediction errors to estimate scalar action values from previous experience, while "mo...

متن کامل

Neural basis of reinforcement learning and decision making.

Reinforcement learning is an adaptive process in which an animal utilizes its previous experience to improve the outcomes of future choices. Computational theories of reinforcement learning play a central role in the newly emerging areas of neuroeconomics and decision neuroscience. In this framework, actions are chosen according to their value functions, which describe how much future reward is...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Electronics

سال: 2021

ISSN: ['2079-9292']

DOI: https://doi.org/10.3390/electronics10030227